Goto

Collaborating Authors

 Wiltshire


MCTED: A Machine-Learning-Ready Dataset for Digital Elevation Model Generation From Mars Imagery

Osadnik, Rafał, Gómez, Pablo, Bohacek, Eleni, Bahia, Rickbir

arXiv.org Artificial Intelligence

This work presents a new dataset for the Martian digital elevation model prediction task, ready for machine learning applications called MCTED. The dataset has been generated using a comprehensive pipeline designed to process high-resolution Mars orthoimage and DEM pairs from Day et al., yielding a dataset consisting of 80,898 data samples. The source images are data gathered by the Mars Reconnaissance Orbiter using the CTX instrument, providing a very diverse and comprehensive coverage of the Martian surface. Given the complexity of the processing pipelines used in large-scale DEMs, there are often artefacts and missing data points in the original data, for which we developed tools to solve or mitigate their impact. We divide the processed samples into training and validation splits, ensuring samples in both splits cover no mutual areas to avoid data leakage. Every sample in the dataset is represented by the optical image patch, DEM patch, and two mask patches, indicating values that were originally missing or were altered by us. This allows future users of the dataset to handle altered elevation regions as they please. We provide statistical insights of the generated dataset, including the spatial distribution of samples, the distributions of elevation values, slopes and more. Finally, we train a small U-Net architecture on the MCTED dataset and compare its performance to a monocular depth estimation foundation model, DepthAnythingV2, on the task of elevation prediction. We find that even a very small architecture trained on this dataset specifically, beats a zero-shot performance of a depth estimation foundation model like DepthAnythingV2. We make the dataset and code used for its generation completely open source in public repositories.


British troops to be given powers to shoot down drones on sight, Telegraph reports

The Japan Times

John Healey, the British defense secretary, tours a new military drone production facility in Swindon, U.K., on Sept. 15. Healey is reportedly set to authorize new powers to shoot down drones amid a rise in incursions. British troops will be given new powers to shoot down drones threatening U.K. military bases, the Telegraph reported on Sunday, citing an upcoming announcement on Monday from John Healey, the British defense secretary. Healey is expected to unveil his vision on how to protect Britain's most critical military bases in response to a growing threat posed by Russia, the newspaper said. Although the new powers will initially apply only for military sites, the British government was not ruling out working to extend those powers to other important sites like airports, the Telegraph said, citing a source.


VILENS: Visual, Inertial, Lidar, and Leg Odometry for All-Terrain Legged Robots

Wisth, David, Camurri, Marco, Fallon, Maurice

arXiv.org Artificial Intelligence

We present visual inertial lidar legged navigation system (VILENS), an odometry system for legged robots based on factor graphs. The key novelty is the tight fusion of four different sensor modalities to achieve reliable operation when the individual sensors would otherwise produce degenerate estimation. To minimize leg odometry drift, we extend the robot's state with a linear velocity bias term, which is estimated online. This bias is observable because of the tight fusion of this preintegrated velocity factor with vision, lidar, and inertial measurement unit (IMU) factors. Extensive experimental validation on different ANYmal quadruped robots is presented, for a total duration of 2 h and 1.8 km traveled. The experiments involved dynamic locomotion over loose rocks, slopes, and mud, which caused challenges such as slippage and terrain deformation. Perceptual challenges included dark and dusty underground caverns, and open and feature-deprived areas. We show an average improvement of 62% translational and 51% rotational errors compared to a state-of-the-art loosely coupled approach. To demonstrate its robustness, VILENS was also integrated with a perceptive controller and a local path planner.


Data-Centric Human Preference Optimization with Rationales

Just, Hoang Anh, Jin, Ming, Sahu, Anit, Phan, Huy, Jia, Ruoxi

arXiv.org Artificial Intelligence

Reinforcement learning from human feedback plays a crucial role in aligning language models towards human preferences, traditionally represented through comparisons between pairs or sets of responses within a given context. While many studies have enhanced algorithmic techniques to optimize learning from such data, this work shifts focus to improving preference learning through a data-centric approach. Specifically, we propose enriching existing preference datasets with machine-generated rationales that explain the reasons behind choices. We develop a simple and principled framework to augment current preference learning methods with rationale information. Our comprehensive analysis highlights how rationales enhance learning efficiency. Extensive experiments reveal that rationale-enriched preference learning offers multiple advantages: it improves data efficiency, accelerates convergence to higher-performing models, and reduces verbosity bias and hallucination. Furthermore, this framework is versatile enough to integrate with various preference optimization algorithms. Overall, our findings highlight the potential of re-imagining data design for preference learning, demonstrating that even freely available machine-generated rationales can significantly boost performance across multiple dimensions. The code repository is available at https: //github.com/reds-lab/preference-learning-with-rationales


LEADER: Learning Attention over Driving Behaviors for Planning under Uncertainty

Danesh, Mohamad H., Cai, Panpan, Hsu, David

arXiv.org Artificial Intelligence

Uncertainty on human behaviors poses a significant challenge to autonomous driving in crowded urban environments. The partially observable Markov decision processes (POMDPs) offer a principled framework for planning under uncertainty, often leveraging Monte Carlo sampling to achieve online performance for complex tasks. However, sampling also raises safety concerns by potentially missing critical events. To address this, we propose a new algorithm, LEarning Attention over Driving bEhavioRs (LEADER), that learns to attend to critical human behaviors during planning. LEADER learns a neural network generator to provide attention over human behaviors in real-time situations. It integrates the attention into a belief-space planner, using importance sampling to bias reasoning towards critical events. To train the algorithm, we let the attention generator and the planner form a min-max game. By solving the min-max game, LEADER learns to perform risk-aware planning without human labeling.


An adaptive music generation architecture for games based on the deep learning Transformer mode

Santos, Gustavo Amaral Costa dos, Baffa, Augusto, Briot, Jean-Pierre, Feijó, Bruno, Furtado, Antonio Luz

arXiv.org Artificial Intelligence

This paper presents an architecture for generating music for video games based on the Transformer deep learning model. Our motivation is to be able to customize the generation according to the taste of the player, who can select a corpus of training examples, corresponding to his preferred musical style. The system generates various musical layers, following the standard layering strategy currently used by composers designing video game music. To adapt the music generated to the game play and to the player(s) situation, we are using an arousal-valence model of emotions, in order to control the selection of musical layers. We discuss current limitations and prospects for the future, such as collaborative and interactive control of the musical components.


How the spirit of ancient Stonehenge was captured with a 21st-century drone

National Geographic

Reuben Wu, a British photographer and visual artist based in Chicago, was first introduced to National Geographic as most people are: When he was a child, he enjoyed looking at the magazines his father subscribed to for decades. He dreamed of seeing his photographs in the same magazine--and even on the cover. So when National Geographic asked him to photograph an iconic monument he knows well, he was ready to work. Last summer, Wu experienced a stark contrast of modern and prehistoric, as he used drones and artificial light to photograph Stonehenge, one of the best-known prehistoric monuments, while hearing honking cars passing by. The site in Wiltshire, England, is bisected by the A303--a major road that may soon be in a tunnel should a 2020 proposal become reality--which means motorists may have seen Wu's photo shoot and lit-up drones.


Towards human-agent knowledge fusion (HAKF) in support of distributed coalition teams

Braines, Dave, Cerutti, Federico, Vilamala, Marc Roig, Srivastava, Mani, Preece, Lance Kaplan Alun, Pearson, Gavin

arXiv.org Artificial Intelligence

Future coalition operations can be substantially augmented through agile teaming between human and machine agents, but in a coalition context these agents may be unfamiliar to the human users and expected to operate in a broad set of scenarios rather than being narrowly defined for particular purposes. In such a setting it is essential that the human agents can rapidly build trust in the machine agents through appropriate transparency of their behaviour, e.g., through explanations. The human agents are also able to bring their local knowledge to the team, observing the situation unfolding and deciding which key information should be communicated to the machine agents to enable them to better account for the particular environment. In this paper we describe the initial steps towards this human-agent knowledge fusion (HAKF) environment through a recap of the key requirements, and an explanation of how these can be fulfilled for an example situation. We show how HAKF has the potential to bring value to both human and machine agents working as part of a distributed coalition team in a complex event processing setting with uncertain sources.


Mediation Challenges and Socio-Technical Gaps for Explainable Deep Learning Applications

Brandão, Rafael, Carbonera, Joel, de Souza, Clarisse, Ferreira, Juliana, Gonçalves, Bernardo, Leitão, Carla

arXiv.org Artificial Intelligence

The presumed data owners' right to explanations brought about by the General Data Protection Regulation in Europe has shed light on the social challenges of explainable artificial intelligence (XAI). In this paper, we present a case study with Deep Learning (DL) experts from a research and development laboratory focused on the delivery of industrial-strength AI technologies. Our aim was to investigate the social meaning (i.e. meaning to others) that DL experts assign to what they do, given a richly contextualized and familiar domain of application. Using qualitative research techniques to collect and analyze empirical data, our study has shown that participating DL experts did not spontaneously engage into considerations about the social meaning of machine learning models that they build. Moreover, when explicitly stimulated to do so, these experts expressed expectations that, with real-world DL application, there will be available mediators to bridge the gap between technical meanings that drive DL work, and social meanings that AI technology users assign to it. We concluded that current research incentives and values guiding the participants' scientific interests and conduct are at odds with those required to face some of the scientific challenges involved in advancing XAI, and thus responding to the alleged data owners' right to explanations or similar societal demands emerging from current debates. As a concrete contribution to mitigate what seems to be a more general problem, we propose three preliminary XAI Mediation Challenges with the potential to bring together technical and social meanings of DL applications, as well as to foster much needed interdisciplinary collaboration among AI and the Social Sciences researchers.


Deploying AI Frameworks on Secure HPC Systems with Containers

Brayford, David, Vallecorsa, Sofia, Atanasov, Atanas, Baruffa, Fabio, Riviera, Walter

arXiv.org Artificial Intelligence

The increasing interest in the usage of Artificial Intelligence techniques (AI) from the research community and industry to tackle "real world" problems, requires High Performance Computing (HPC) resources to efficiently compute and scale complex algorithms across thousands of nodes. Unfortunately, typical data scientists are not familiar with the unique requirements and characteristics of HPC environments. They usually develop their applications with high-level scripting languages or frameworks such as TensorFlow and the installation process often requires connection to external systems to download open source software during the build. HPC environments, on the other hand, are often based on closed source applications that incorporate parallel and distributed computing API's such as MPI and OpenMP, while users have restricted administrator privileges, and face security restrictions such as not allowing access to external systems. In this paper we discuss the issues associated with the deployment of AI frameworks in a secure HPC environment and how we successfully deploy AI frameworks on SuperMUC-NG with Charliecloud.